Main
Sneha Kannan
I am adept at processing large quantities of complex data, feature engineering, building and deploying machine learning models in production by collaborating with product managers, data engineers and software developers. I am seeking a new opportunity to utilize my skills and broaden my experience while I contribute to the growth of the organization.
Work Experience
Data Scientist II
Braviant Holdings
Chicago, IL
2020 Mar
- Analyzed employer names from the loan applications by fuzzy matching and categorized them into common industry types to understand the impact of the current economic scenario on the product
Senior Analyst
Portfolio & Fraud Analytics, Enova International
Chicago, IL
2020 Feb - 2017 July
- Reduced non-payment rate by 300 bps with a potential saving of 800K R$/year by building a GBM model using transactions from Bank Statements and deploying it in AWS Sagemaker
- Analyzed the profitability of funded loans and helped determine budgeting targets for the portfolio by assessing the Net Present Value incorporating marketing costs, underwriting and operational costs
- Reduced funding fraud loan applications and potentially saved 1.1M R$/year by building a fraud tracking database with advanced reporting and real time dashboard using web app framework Flask Python
- Increased productivity for fraud operations team by 25% using a recommendation model to generate a similarity score between all loan applications and known fraud loans
- Conducted various ad-hoc analysis by using complex PostgreSQL queries and statistical techniques to improve on the existing risk management, marketing and operational strategies
Analyst
Portfolio Analytics, Enova International
Chicago, IL
2017 June - 2015 July
- Increased the recovery rate from defaulted loans by 11% by developing a collections model to predict the likelihood of payment and optimized the debt collections process
- Improved conversions and reduced default rate by building a model suite to carry out A/B tests during loan processing while reducing risk to the product and lowering friction to the end users
- Reduced the underwriting cost by 250K R$/year by optimizing the underwriting process with the help of a machine learning model and ensured better customer segmentation
- Built a statistical model facilitating the direct mail campaign for a product which provided a 23% lift in KS and increased the AUC by 0.04 compared to the existing model
Education
Master’s in Management Information Systems
University of Illinois at Chicago
Chicago, IL
2015 May - 2013 Aug
Graduate Projects
Exploring jobs on Analytics: Natural Language Processing (NLP)
N/A
N/A
N/A
- Identified different features of jobs related to Analytics & Big data by collecting data from a popular job listings website. Discovered word clusters using the technique of topic modeling on the job descriptions
Item-based collaborative filtering for movie recommendation
N/A
N/A
N/A
- Developed a MapReduce program in Hadoop to compute Pearson correlation similarity for all pairs of movies in the data set. Used this to compute the prediction on a movie for a given user using the weighted sum technique and recommended top k movies to the user
Frequent itemset mining using MapReduce
N/A
N/A
N/A
- Generated 2 and 3-item sets using MapReduce from a large transaction data set using association rules. Identified items which were frequently bought together and those which were bought independently
Clustering using K-means under Apache Mahout
N/A
N/A
N/A
- Preprocessed and cleaned the JSON input files of the Yelp reviews data set. Applied K-means clustering under Mahout and Hadoop environment to find clusters in the reviews
Strategic Operating Plan to a group of Department stores
N/A
N/A
N/A
- Used the SQL Server Analysis Services Cubes for analyzing the stores data available through the University of Arkansas website. Created a report which identified the stores which were least profitable using the profitability KPI metrics, days on which the stores were least profitable so they can reduce the operations on those days and least profitable brands in the stores so their inventory can be reduced